698 research outputs found

    Sex-specific fundamental and formant frequency patterns in a cross-sectional study

    Get PDF
    An extensive developmental acoustic study of the speech patterns of children and adults was reported by Lee and colleagues [Lee et al., J. Acoust. Soc. Am. 105, 1455-1468 (1999)]. This paper presents a reexamination of selected fundamental frequency and formant frequency data presented in their report for 10 monophthongs by investigating sex-specific and developmental patterns using two different approaches. The first of these includes the investigation of age- and sex-specific formant frequency patterns in the monophthongs. The second, the investigation of fundamental frequency and formant frequency data using the critical band rate (bark) scale and a number of acoustic-phonetic dimensions of the monophthongs from an age- and sex-specific perspective. These acoustic-phonetic dimensions include: vowel spaces and distances from speaker centroids; frequency differences between the formant frequencies of males and females; vowel openness/closeness and frontness/backness; the degree of vocal effort; and formant frequency ranges. Both approaches reveal both age- and sex-specific development patterns which also appear to be dependent on whether vowels are peripheral or non-peripheral. The developmental emergence of these sex-specific differences are discussed with reference to anatomical, physiological, sociophonetic and culturally determined factors. Some directions for further investigation into the age-linked sex differences in speech across the lifespan are also proposed

    A typology of laterals in twelve English dialects

    Get PDF
    Allophonic patterns of variation in English laterals have been well studied in phonetics and phonology for decades, but establishing broad generalizations across varieties has proven challenging. In this study, we advance a typology of onset/coda lateral distinctions in English, using crowdsourced recordings from 95 speakers across twelve dialects of Anglo (UK) English. Results confirm the existence of dialects with and without onset/coda distinctions, and conditional inference trees are used to identity three main patterns in the data: (1) clear onsets and dark codas; (2) intermediate/dark onsets and dark codas, but with a positional distinction intact; (3) dark onsets and dark codas, with minimal or no distinctions between positions

    Automatic recognition of schwa variants in spontaneous Hungarian speech

    Get PDF
    This paper analyzes the nature of the process involved in optional vowel reduction in Hungarian, and the acoustic structure of schwa variants in spontaneous speech. The study focuses on the acoustic patterns of both the basic realizations of Hungarian vowels and their realizations as neutral vowels (schwas), as well as on the design, implementation, and evaluation of a set of algorithms for the recognition of both types of realizations from the speech waveform. The authors address the question whether schwas form a unified group of vowels or they show some dependence on the originally intended articulation of the vowel they stand for. The acoustic study uses a database consisting of over 4,000 utterances extracted from continuous speech, and recorded from 19 speakers. The authors propose methods for the recognition of neutral vowels depending on the various vowels they replace in spontaneous speech. Mel-Frequency Cepstral Coefficients are calculated and used for the training of Hidden Markov Models. The recognition system was trained on 2,500 utterances and then tested on 1,500 utterances. The results show that a neutral vowel can be detected in 72% of all occurrences. Stressed and unstressed syllables can be distinguished in 92% of all cases. Neutralized vowels do not form a unified group of phoneme realizations. The pronunciation of schwa heavily depends on the original articulation configuration of the intended vowel

    The Role of Native Language and the Fundamental Design of the Auditory System in Detecting Rhythm Changes

    Get PDF
    Accepted December 13, 2018Purpose: We investigated whether rhythm discrimination is mainly driven by the native language of the listener or by the fundamental design of the human auditory system and universal cognitive mechanisms shared by all people irrespective of rhythmic patterns in their native language. Method: In multiple experiments, we asked participants to listen to 2 continuous acoustic sequences and to determine whether their rhythms were the same or different (AX discrimination). Participants were native speakers of 4 languages with different rhythmic properties (Spanish, French, English, and German) to understand whether the predominant rhythmic patterns of a native language affect sensitivity, bias, and reaction time in detecting rhythmic changes in linguistic (Experiment 2) and in nonlinguistic (Experiments 1 and 2) acoustic sequences. We examined sensitivity and bias measures, as well as reaction times. We also computed Bayes factors in order to assess the effect of native language. Results: All listeners performed better (i.e., responded faster and manifested higher sensitivity and accuracy) when detecting the presence or absence of a rhythm change when the 1st stimulus in an AX test pair exhibited regular rhythm (i.e., a syllable-timed rhythmic pattern) than when the 1st stimulus exhibited irregular rhythm (i.e., stress-timed rhythmic pattern). This result pattern was observed both on linguistic and nonlinguistic stimuli and was not modulated by the native language of the participant. Conclusion: We conclude that rhythm change detection is a fundamental function of a processing system that relies on general auditory mechanisms and is not modulated by linguistic experience.The authors acknowledge support from Spanish Ministry of Economy and Competitiveness Grant PSI2017-82563-P (awarded to A. G. S.), the “Severo Ochoa” Programme for Centres/Units of Excellence in R&D Grant SEV-2015-490 (BCBL), and the Basque Foundation for Science Grant IKERBASQUE (awarded to A. G. S. and M. O.). D. M. G. was supported by Grant PIA/Basal FB0003 from the Chilean Research Council. L. P. was supported by the Spanish Ministry of Economy and Competitiveness via Juan de la Cierva fellowship

    Comparing phoneme frequency, age of acquisition, and loss in aphasia:Implications for phonological universals

    Get PDF
    Phonological complexity may be central to the nature of human language. It may shape the distribution of phonemes and phoneme sequences within languages, but also determine age of acquisition and susceptibility to loss in aphasia. We evaluated this claim using frequency statistics derived from a corpus of phonologically transcribed Italian words (phonitalia, available at phonitalia,org), rankings of phoneme age of acquisition (AoA) and rate of phoneme errors in patients with apraxia of speech (AoS) as an indication of articulatory complexity. These measures were related to cross-linguistically derived markedness rankings. We found strong correspondences. AoA, however, was predicted by both apraxic errors and frequency, suggesting independent contributions of these variables. Our results support the reality of universal principles of complexity. In addition they suggest that these complexity principles have articulatory underpinnings since they modulate the production of patients with AoS, but not the production of patients with more central phonological difficulties

    Zebra finches and Dutch adults exhibit the same cue weighting bias in vowel perception

    Get PDF
    Vocal tract resonances, called formants, are the most important parameters in human speech production and perception. They encode linguistic meaning and have been shown to be perceived by a wide range of species. Songbirds are also sensitive to different formant patterns in human speech. They can categorize words differing only in their vowels based on the formant patterns independent of speaker identity in a way comparable to humans. These results indicate that speech perception mechanisms are more similar between songbirds and humans than realized before. One of the major questions regarding formant perception concerns the weighting of different formants in the speech signal (“acoustic cue weighting”) and whether this process is unique to humans. Using an operant Go/NoGo design, we trained zebra finches to discriminate syllables, whose vowels differed in their first three formants. When subsequently tested with novel vowels, similar in either their first formant or their second and third formants to the familiar vowels, similarity in the higher formants was weighted much more strongly than similarity in the lower formant. Thus, zebra finches indeed exhibit a cue weighting bias. Interestingly, we also found that Dutch speakers when tested with the same paradigm exhibit the same cue weighting bias. This, together with earlier findings, supports the hypothesis that human speech evolution might have exploited general properties of the vertebrate auditory system
    • …
    corecore